# Local Window Attention
Swinv2 Base Patch4 Window12 192 22k
Apache-2.0
Swin Transformer v2 is a vision Transformer model that achieves efficient image processing through hierarchical feature maps and local window self-attention mechanisms.
Image Classification
Transformers

S
microsoft
8,603
3
Swin Large Patch4 Window12 384
Apache-2.0
Swin Transformer is a hierarchical vision Transformer model based on shifted windows, specifically designed for image classification tasks.
Image Classification
Transformers

S
microsoft
22.77k
1
Swin Base Patch4 Window7 224 In22k
Apache-2.0
Swin Transformer is a hierarchical window-based vision Transformer model pretrained on the ImageNet-21k dataset, suitable for image classification tasks.
Image Classification
Transformers

S
microsoft
13.30k
15
Swin Small Patch4 Window7 224
Apache-2.0
Swin Transformer is a hierarchical window-based vision Transformer model designed for image classification tasks, with computational complexity linearly related to input image size.
Image Classification
Transformers

S
microsoft
2,028
1
Swin Large Patch4 Window12 384 In22k
Apache-2.0
Swin Transformer is a hierarchical window-based vision Transformer model, pretrained on the ImageNet-21k dataset, suitable for image classification tasks.
Image Classification
Transformers

S
microsoft
1,063
7
Swin Tiny Patch4 Window7 224
Apache-2.0
Swin Transformer is a hierarchical vision Transformer that achieves linear computational complexity by computing self-attention within local windows, making it suitable for image classification tasks.
Image Classification
Transformers

S
microsoft
98.00k
42
Swin Large Patch4 Window7 224
Apache-2.0
Swin Transformer is a hierarchical vision Transformer that achieves linear computational complexity by computing self-attention within local windows, making it suitable for image classification and dense recognition tasks.
Image Classification
Transformers

S
microsoft
2,079
1
Featured Recommended AI Models